# GPTQ quantization
Qwen3 Embedding 4B W4A16 G128
Apache-2.0
This is the Qwen3-Embedding-4B model after GPTQ quantization, with significantly reduced video memory usage and minimal performance loss.
Text Embedding
Q
boboliu
141
1
Qwen3 Embedding 8B W4A16 G128
Apache-2.0
GPTQ quantized version of Qwen3-Embedding-8B, significantly reducing VRAM requirements while maintaining high performance
Text Embedding
Q
boboliu
322
1
Featured Recommended AI Models